Goto

Collaborating Authors

 inducing brain-relevant bias


Inducing brain-relevant bias in natural language processing models

Neural Information Processing Systems

Progress in natural language processing (NLP) models that estimate representations of word sequences has recently been leveraged to improve the understanding of language processing in the brain. However, these models have not been specifically designed to capture the way the brain represents language meaning. We hypothesize that fine-tuning these models to predict recordings of brain activity of people reading text will lead to representations that encode more brain-activity-relevant language information. We demonstrate that a version of BERT, a recently introduced and powerful language model, can improve the prediction of brain activity after fine-tuning. We show that the relationship between language and brain activity learned by BERT during this fine-tuning transfers across multiple participants. We also show that, for some participants, the fine-tuned representations learned from both magnetoencephalography (MEG) and functional magnetic resonance imaging (fMRI) are better for predicting fMRI than the representations learned from fMRI alone, indicating that the learned representations capture brain-activity-relevant information that is not simply an artifact of the modality. While changes to language representations help the model predict brain activity, they also do not harm the model's ability to perform downstream NLP tasks. Our findings are notable for research on language understanding in the brain.



Reviews: Inducing brain-relevant bias in natural language processing models

Neural Information Processing Systems

The reviewers agree that this paper makes an original contribution in investigating the fine tuning of a contextual embedding model, e.g. They have identified some issues with the clarity of the motivation of the work and the presentation of some of the results, but I feel that these shortcomings are outweighed by the contributions of this work which should be of interest to researchers across a number of disciplines. The reviewers thank the authors for the additional analysis provided in their response and look forward to this being incorporated into the final paper. In particular it would be good to clarify to significance of the results in Table 1, whether any of the differences are interpretable or whether this simply shows that they are equivalent.


Inducing brain-relevant bias in natural language processing models

Neural Information Processing Systems

Progress in natural language processing (NLP) models that estimate representations of word sequences has recently been leveraged to improve the understanding of language processing in the brain. However, these models have not been specifically designed to capture the way the brain represents language meaning. We hypothesize that fine-tuning these models to predict recordings of brain activity of people reading text will lead to representations that encode more brain-activity-relevant language information. We demonstrate that a version of BERT, a recently introduced and powerful language model, can improve the prediction of brain activity after fine-tuning. We show that the relationship between language and brain activity learned by BERT during this fine-tuning transfers across multiple participants.


Inducing brain-relevant bias in natural language processing models

Schwartz, Dan, Toneva, Mariya, Wehbe, Leila

Neural Information Processing Systems

Progress in natural language processing (NLP) models that estimate representations of word sequences has recently been leveraged to improve the understanding of language processing in the brain. However, these models have not been specifically designed to capture the way the brain represents language meaning. We hypothesize that fine-tuning these models to predict recordings of brain activity of people reading text will lead to representations that encode more brain-activity-relevant language information. We demonstrate that a version of BERT, a recently introduced and powerful language model, can improve the prediction of brain activity after fine-tuning. We show that the relationship between language and brain activity learned by BERT during this fine-tuning transfers across multiple participants.